Goto

Collaborating Authors

 Capital Governorate


Cross-Domain Generalization of Multimodal LLMs for Global Photovoltaic Assessment

Guo, Muhao, Weng, Yang

arXiv.org Artificial Intelligence

Table I summarizes the datasets used for training and evaluation. Both baseline models and the PV AL framework were fine-tuned on 2,000 annotated tiles from Santa Ana, CA. The large-scale evaluation set includes about 100,000 tiles from Tempe and Santa Ana, while 480 tiles per region were used for cross-domain generalization tests across diverse climates and geographies. B. Multimodal LLM Configuration Configuring the PV AL system for solar panel detection involves a multi-faceted approach that integrates prompt engineering, output standardization, and supervised fine-tuning. This configuration is critical for steering the foundational GPT -4o model towards the specific, high-precision task of geospatial analysis. Prompt Task Decomposition Identify the presence of solar panels in images of residential rooftops, and determine their locations and quantity within the images. You will be provided with images that may contain residential rooftop solar systems. Analyze each image to detect solar panels. Steps: 1. ** Image Analysis **: Examine the entire image to identify any objects that appear to be solar panels.



DRAGON: Guard LLM Unlearning in Context via Negative Detection and Reasoning

Wang, Yaxuan, Liu, Chris Yuhao, Liu, Quan, Pang, Jinglong, Wei, Wei, Bao, Yujia, Liu, Yang

arXiv.org Artificial Intelligence

Unlearning in Large Language Models (LLMs) is crucial for protecting private data and removing harmful knowledge. Most existing approaches rely on fine-tuning to balance unlearning efficiency with general language capabilities. However, these methods typically require training or access to retain data, which is often unavailable in real world scenarios. Although these methods can perform well when both forget and retain data are available, few works have demonstrated equivalent capability in more practical, data-limited scenarios. To overcome these limitations, we propose Detect-Reasoning Augmented GeneratiON (DRAGON), a systematic, reasoning-based framework that utilizes in-context chain-of-thought (CoT) instructions to guard deployed LLMs before inference. Instead of modifying the base model, DRAGON leverages the inherent instruction-following ability of LLMs and introduces a lightweight detection module to identify forget-worthy prompts without any retain data. These are then routed through a dedicated CoT guard model to enforce safe and accurate in-context intervention. To robustly evaluate unlearning performance, we introduce novel metrics for unlearning performance and the continual unlearning setting. Extensive experiments across three representative unlearning tasks validate the effectiveness of DRAGON, demonstrating its strong unlearning capability, scalability, and applicability in practical scenarios.


Oblivionis: A Lightweight Learning and Unlearning Framework for Federated Large Language Models

Zhang, Fuyao, Yan, Xinyu, Wu, Tiantong, Li, Wenjie, Chen, Tianxiang, Cao, Yang, Yan, Ran, Huang, Longtao, Lim, Wei Yang Bryan, Yang, Qiang

arXiv.org Artificial Intelligence

Large Language Models (LLMs) increasingly leverage Federated Learning (FL) to utilize private, task-specific datasets for fine-tuning while preserving data privacy. However, while federated LLM frameworks effectively enable collaborative training without raw data sharing, they critically lack built-in mechanisms for regulatory compliance like GDPR's right to be forgotten. Integrating private data heightens concerns over data quality and long-term governance, yet existing distributed training frameworks offer no principled way to selectively remove specific client contributions post-training. Due to distributed data silos, stringent privacy constraints, and the intricacies of interdependent model aggregation, federated LLM unlearning is significantly more complex than centralized LLM unlearning. T o address this gap, we introduce Oblivionis, a lightweight learning and unlearning framework that enables clients to selectively remove specific private data during federated LLM training, enhancing trustworthiness and regulatory compliance. By unifying FL and unlearning as a dual optimization objective, we incorporate 6 FL and 5 unlearning algorithms for comprehensive evaluation and comparative analysis, establishing a robust pipeline for federated LLM unlearning. Extensive experiments demonstrate that Oblivionis outperforms local training, achieving a robust balance between forgetting efficacy and model utility, with cross-algorithm comparisons providing clear directions for future LLM development.


Label Smoothing Improves Gradient Ascent in LLM Unlearning

Pang, Zirui, Zheng, Hao, Deng, Zhijie, Li, Ling, Zhong, Zixin, Wei, Jiaheng

arXiv.org Artificial Intelligence

LLM unlearning has emerged as a promising approach, aiming to enable models to forget hazardous/undesired knowledge at low cost while preserving as much model utility as possible. Among existing techniques, the most straightforward method is performing Gradient Ascent (GA) w.r.t. the forget data, thereby forcing the model to unlearn the forget dataset. However, GA suffers from severe instability, as it drives updates in a divergent direction, often resulting in drastically degraded model utility. To address this issue, we propose Smoothed Gradient Ascent (SGA). SGA combines the forget data with multiple constructed normal data through a tunable smoothing rate. Intuitively, this extends GA from learning solely on the forget data to jointly learning across both forget and normal data, enabling more stable unlearning while better preserving model utility. Theoretically, we provide the theoretical guidance on the selection of the optimal smoothing rate. Empirically, we evaluate SGA on three benchmarks: TOFU, Harry Potter, and MUSE-NEWS. Experimental results demonstrate that SGA consistently outperforms the original Gradient Ascent (GA) method across all metrics and achieves top-2 performance among all baseline methods on several key metrics. Since such knowledge is embedded in model representations, it can easily surface in outputs.


Constrained Entropic Unlearning: A Primal-Dual Framework for Large Language Models

Entesari, Taha, Hatami, Arman, Khaziev, Rinat, Ramakrishna, Anil, Fazlyab, Mahyar

arXiv.org Artificial Intelligence

Large Language Models (LLMs) deployed in real-world settings increasingly face the need to unlearn sensitive, outdated, or proprietary information. Existing unlearning methods typically formulate forgetting and retention as a regularized trade-off, combining both objectives into a single scalarized loss. This often leads to unstable optimization and degraded performance on retained data, especially under aggressive forgetting. We propose a new formulation of LLM unlearning as a constrained optimization problem: forgetting is enforced via a novel logit-margin flattening loss that explicitly drives the output distribution toward uniformity on a designated forget set, while retention is preserved through a hard constraint on a separate retain set. Compared to entropy-based objectives, our loss is softmax-free, numerically stable, and maintains non-vanishing gradients, enabling more efficient and robust optimization. We solve the constrained problem using a scalable primal-dual algorithm that exposes the trade-off between forgetting and retention through the dynamics of the dual variable, all without any extra computational overhead. Evaluations on the TOFU and MUSE benchmarks across diverse LLM architectures demonstrate that our approach consistently matches or exceeds state-of-the-art baselines, effectively removing targeted information while preserving downstream utility.


Forget to Know, Remember to Use: Context-Aware Unlearning for Large Language Models

Peng, Yuefeng, Afshar, Parnian, Ganji, Megan, Butler, Thomas, Houmansadr, Amir, Wang, Mingxian, Hong, Dezhi

arXiv.org Artificial Intelligence

Large language models may encode sensitive information or outdated knowledge that needs to be removed, to ensure responsible and compliant model responses. Unlearning has emerged as an efficient alternative to full retraining, aiming to remove specific knowledge while preserving overall model utility. Existing evaluations of unlearning methods focus on (1) the extent of forgetting of the target knowledge (forget set) and (2) maintaining performance on the retain set (i.e., utility). However, these evaluations overlook an important usability aspect: users may still want the model to leverage the removed information if it is re-introduced in the prompt. In a systematic evaluation of six state-of-the-art unlearning methods, we find that they consistently impair such contextual utility. To address this, we augment unlearning objectives with a plug-in term that preserves the model's ability to use forgotten knowledge when it is present in context. Extensive experiments demonstrate that our approach restores contextual utility to near original levels while still maintaining effective forgetting and retain-set utility. Large language models (LLMs) (Y ang et al., 2025a; Team et al., 2024; Dubey et al., 2024) are trained on massive web-scale datasets that can unintentionally include sensitive or outdated information (Henderson et al., 2023; Li et al., 2024; Carlini et al., 2021; Nasr et al., 2025). Such information may later need to be removed to ensure responsible and reliable model behavior. A straightforward solution is to remove the targeted data (the forget set) from the training data and retrain the model. However, retraining billion-parameter-scale LLMs is prohibitively costly and time-consuming.


Reference-Specific Unlearning Metrics Can Hide the Truth: A Reality Check

Cho, Sungjun, Hwang, Dasol, Sala, Frederic, Hwang, Sangheum, Cho, Kyunghyun, Cha, Sungmin

arXiv.org Artificial Intelligence

Current unlearning metrics for generative models evaluate success based on reference responses or classifier outputs rather than assessing the core objective: whether the unlearned model behaves indistinguishably from a model that never saw the unwanted data. This reference-specific approach creates systematic blind spots, allowing models to appear successful while retaining unwanted knowledge accessible through alternative prompts or attacks. We address these limitations by proposing Functional Alignment for Distributional Equivalence (FADE), a novel metric that measures distributional similarity between unlearned and reference models by comparing bidirectional likelihood assignments over generated samples. Unlike existing approaches that rely on predetermined references, FADE captures functional alignment across the entire output distribution, providing a principled assessment of genuine unlearning. Our experiments on the TOFU benchmark for LLM unlearning and the UnlearnCanvas benchmark for text-to-image diffusion model unlearning reveal that methods achieving near-optimal scores on traditional metrics fail to achieve distributional equivalence, with many becoming more distant from the gold standard than before unlearning. These findings expose fundamental gaps in current evaluation practices and demonstrate that FADE provides a more robust foundation for developing and assessing truly effective unlearning methods.



Solar Photovoltaic Assessment with Large Language Model

Guo, Muhao, Weng, Yang

arXiv.org Artificial Intelligence

Accurate detection and localization of solar photovoltaic (PV) panels in satellite imagery is essential for optimizing microgrids and active distribution networks (ADNs), which are critical components of renewable energy systems. Existing methods lack transparency regarding their underlying algorithms or training datasets, rely on large, high-quality PV training data, and struggle to generalize to new geographic regions or varied environmental conditions without extensive re-training. These limitations lead to inconsistent detection outcomes, hindering large-scale deployment and data-driven grid optimization. In this paper, we investigate how large language models (LLMs) can be leveraged to overcome these challenges. Despite their promise, LLMs face several challenges in solar panel detection, including difficulties with multi-step logical processes, inconsistent output formatting, frequent misclassification of visually similar objects (e.g., shadows, parking lots), and low accuracy in complex tasks such as spatial localization and quantification. To overcome these issues, we propose the PV Assessment with LLMs (PVAL) framework, which incorporates task decomposition for more efficient workflows, output standardization for consistent and scalable formatting, few-shot prompting to enhance classification accuracy, and fine-tuning using curated PV datasets with detailed annotations. PVAL ensures transparency, scalability, and adaptability across heterogeneous datasets while minimizing computational overhead. By combining open-source accessibility with robust methodologies, PVAL establishes an automated and reproducible pipeline for solar panel detection, paving the way for large-scale renewable energy integration and optimized grid management.